电线杆和建筑物边缘经常是城市道路上可观察到的对象,为各种计算机视觉任务提供了可靠的提示。为了重复提取它们作为特征并在离散激光镜头框架之间进行注册,我们提出了第一个基于学习的功能分割和LIDAR点云中3D线的描述模型。为了训练我们的模型,而无需耗时和乏味的数据标记过程,我们首先生成了目标线基本外观的合成原始图,并构建一个迭代线自动标记的过程,以逐步完善真实激光扫描的线路标签。我们的分割模型可以在任意规模的扰动下提取线,我们使用共享的EDGECONV编码层共同训练两个分割和描述符头。基于模型,我们可以在没有初始转换提示的情况下构建一个高度可用的全局注册模块,用于点云注册。实验表明,我们基于线的注册方法对基于最先进的方法的方法具有很高的竞争力。我们的代码可在https://github.com/zxrzju/superline3d.git上找到。
translated by 谷歌翻译
在riemannian歧管中,Ricci流是用于发展度量的部分微分方程,以便更加规则。我们希望来自此类指标的拓扑结构可用于帮助机器学习的任务。然而,这部分工作仍然缺失。在本文中,我们通过动态稳定的Poincar eMinddings来弥合Ricci流和深神经网络之间的这种差距。结果,我们证明,如果初始指标有$ L ^ 2 $ -norm扰动,它偏离了Poincar \'E球上的双曲度量,这种度量的缩放RICCI-DECurck流程平滑,并将其归因于双曲测量。具体地,Ricci流的作用是用作稳定的Poincar的EAll自然地发展,然后将被映射回欧几里德空间。对于在RICCI流下的这种动态稳定的神经歧管中,嵌入这种歧管的神经网络的收敛性不易受到扰动。我们表明,这种RICCI流动辅助神经网络与其在图像分类任务(CIFAR数据集)上的所有欧几里德版本胜过。
translated by 谷歌翻译
Conventional representation learning algorithms for knowledge graphs (KG) map each entity to a unique embedding vector, ignoring the rich information contained in the neighborhood. We propose a method named StarGraph, which gives a novel way to utilize the neighborhood information for large-scale knowledge graphs to obtain entity representations. An incomplete two-hop neighborhood subgraph for each target node is at first generated, then processed by a modified self-attention network to obtain the entity representation, which is used to replace the entity embedding in conventional methods. We achieved SOTA performance on ogbl-wikikg2 and got competitive results on fb15k-237. The experimental results proves that StarGraph is efficient in parameters, and the improvement made on ogbl-wikikg2 demonstrates its great effectiveness of representation learning on large-scale knowledge graphs. The code is now available at \url{https://github.com/hzli-ucas/StarGraph}.
translated by 谷歌翻译
主动学习(AL)是应选择的数据用于注释。现有的工作试图选择高度不确定或信息性的注释数据。尽管如此,它仍然不清楚所选择的数据如何影响AL中使用的任务模型的测试性能。在这项工作中,我们通过理论上证明,选择更高梯度规范的未标记数据导致测试损失的较低的上限,从而探讨了这种影响,从而产生更好的测试性能。但是,由于缺乏标签信息,直接计算未标记数据的梯度标准是不可行的。为了解决这一挑战,我们提出了两种计划,即预期的Gradnorm和熵 - Gradnorm。前者通过构建预期的经验损失来计算梯度规范,而后者用熵构造无监督的损失。此外,我们将这两个方案集成在通用AL框架中。我们在古典图像分类和语义分割任务中评估我们的方法。为了展示其域应用程序的能力及其对噪声的鲁棒性,我们还在蜂窝成像分析任务中验证了我们的方法,即Cryo-Collecton Subtom图分类。结果表明,我们的方法达到了最先进的卓越性能。我们的源代码可在https://github.com/xulabs/aitom提供
translated by 谷歌翻译
时尚预测学习是给定一系列历史框架的未来框架。传统算法主要基于经常性的神经网络(RNN)。然而,由于经常性结构的序列性,RNN遭受了重大计算负担,例如由于经常性结构的序列性而达到时间和长的背部传播过程。最近,还以编码器 - 解码器或普通编码器的形式研究了基于变压器的方法,但是编码器 - 解码器形式需要过于深的网络,并且普通编码器缺乏短期依赖性。为了解决这些问题,我们提出了一种名为3D时间卷积变压器(TCTN)的算法,其中采用具有时间卷积层的基于变压器的编码器来捕获短期和长期依赖性。由于变压器的并行机理,我们所提出的算法与基于RNN的方法相比,易于实施和培训得多。为了验证我们的算法,我们对移动和kth数据集进行实验,并表明TCTN在性能和训练速度下表现出最先进的(SOTA)方法。
translated by 谷歌翻译
冷冻电子断层扫描(Cryo-et)是一种3D成像技术,可以在近原子分辨率下原位地置于亚细胞结构。细胞冷冻剂图像有助于解决大分子的结构并在单个细胞中确定它们的空间关系,这对细胞和结构生物学具有广泛的意义。体摩数分类和识别构成了这些大分子结构的系统恢复的主要步骤。已被证明监督深度学习方法对重组分类进行高度准确和高效,而是由于缺乏注释数据而受到有限的适用性。虽然生成用于训练监督模型的模拟数据是潜在的解决方案,但与真实实验数据相比,生成数据中的图像强度分布的相当差异将导致训练有素的模型在预测真实错误谱图上预测类别中的差。在这项工作中,我们呈现了低温,一个完全无监督的域适应和随机化框架,用于深入学习的跨域重组分类。我们使用无监督的多逆境域适应来减少模拟和实验数据的特征之间的域移位。我们使用“翘曲”模块开发网络驱动的域随机化过程,以改变模拟数据,并帮助分类器在实验数据上更好地推广。我们不使用任何标记的实验数据来训练我们的模型,而一些现有的替代方法需要标记为跨域分类的实验样本。然而,在本文在本文中,使用两种模拟和实验数据在本文中显示的广泛评估研究中的横域重组分类中现有的替代方法的优先效果优异。
translated by 谷歌翻译
Compressed videos often exhibit visually annoying artifacts, known as Perceivable Encoding Artifacts (PEAs), which dramatically degrade video visual quality. Subjective and objective measures capable of identifying and quantifying various types of PEAs are critical in improving visual quality. In this paper, we investigate the influence of four spatial PEAs (i.e. blurring, blocking, bleeding, and ringing) and two temporal PEAs (i.e. flickering and floating) on video quality. For spatial artifacts, we propose a visual saliency model with a low computational cost and higher consistency with human visual perception. In terms of temporal artifacts, self-attention based TimeSFormer is improved to detect temporal artifacts. Based on the six types of PEAs, a quality metric called Saliency-Aware Spatio-Temporal Artifacts Measurement (SSTAM) is proposed. Experimental results demonstrate that the proposed method outperforms state-of-the-art metrics. We believe that SSTAM will be beneficial for optimizing video coding techniques.
translated by 谷歌翻译
As one of the most important psychic stress reactions, micro-expressions (MEs), are spontaneous and transient facial expressions that can reveal the genuine emotions of human beings. Thus, recognizing MEs (MER) automatically is becoming increasingly crucial in the field of affective computing, and provides essential technical support in lie detection, psychological analysis and other areas. However, the lack of abundant ME data seriously restricts the development of cutting-edge data-driven MER models. Despite the recent efforts of several spontaneous ME datasets to alleviate this problem, it is still a tiny amount of work. To solve the problem of ME data hunger, we construct a dynamic spontaneous ME dataset with the largest current ME data scale, called DFME (Dynamic Facial Micro-expressions), which includes 7,526 well-labeled ME videos induced by 671 participants and annotated by more than 20 annotators throughout three years. Afterwards, we adopt four classical spatiotemporal feature learning models on DFME to perform MER experiments to objectively verify the validity of DFME dataset. In addition, we explore different solutions to the class imbalance and key-frame sequence sampling problems in dynamic MER respectively on DFME, so as to provide a valuable reference for future research. The comprehensive experimental results show that our DFME dataset can facilitate the research of automatic MER, and provide a new benchmark for MER. DFME will be published via https://mea-lab-421.github.io.
translated by 谷歌翻译
Face Anti-spoofing (FAS) is essential to secure face recognition systems from various physical attacks. However, recent research generally focuses on short-distance applications (i.e., phone unlocking) while lacking consideration of long-distance scenes (i.e., surveillance security checks). In order to promote relevant research and fill this gap in the community, we collect a large-scale Surveillance High-Fidelity Mask (SuHiFiMask) dataset captured under 40 surveillance scenes, which has 101 subjects from different age groups with 232 3D attacks (high-fidelity masks), 200 2D attacks (posters, portraits, and screens), and 2 adversarial attacks. In this scene, low image resolution and noise interference are new challenges faced in surveillance FAS. Together with the SuHiFiMask dataset, we propose a Contrastive Quality-Invariance Learning (CQIL) network to alleviate the performance degradation caused by image quality from three aspects: (1) An Image Quality Variable module (IQV) is introduced to recover image information associated with discrimination by combining the super-resolution network. (2) Using generated sample pairs to simulate quality variance distributions to help contrastive learning strategies obtain robust feature representation under quality variation. (3) A Separate Quality Network (SQN) is designed to learn discriminative features independent of image quality. Finally, a large number of experiments verify the quality of the SuHiFiMask dataset and the superiority of the proposed CQIL.
translated by 谷歌翻译
Interview has been regarded as one of the most crucial step for recruitment. To fully prepare for the interview with the recruiters, job seekers usually practice with mock interviews between each other. However, such a mock interview with peers is generally far away from the real interview experience: the mock interviewers are not guaranteed to be professional and are not likely to behave like a real interviewer. Due to the rapid growth of online recruitment in recent years, recruiters tend to have online interviews, which makes it possible to collect real interview data from real interviewers. In this paper, we propose a novel application named EZInterviewer, which aims to learn from the online interview data and provides mock interview services to the job seekers. The task is challenging in two ways: (1) the interview data are now available but still of low-resource; (2) to generate meaningful and relevant interview dialogs requires thorough understanding of both resumes and job descriptions. To address the low-resource challenge, EZInterviewer is trained on a very small set of interview dialogs. The key idea is to reduce the number of parameters that rely on interview dialogs by disentangling the knowledge selector and dialog generator so that most parameters can be trained with ungrounded dialogs as well as the resume data that are not low-resource. Evaluation results on a real-world job interview dialog dataset indicate that we achieve promising results to generate mock interviews. With the help of EZInterviewer, we hope to make mock interview practice become easier for job seekers.
translated by 谷歌翻译